Convergence of online gradient methods for continuous perceptrons with linearly separable training patterns

نویسندگان

  • Wei Wu
  • Zhiqiong Shao
چکیده

h this paper, we prove that the online gradient method for continuous perceptrons converges in finite steps when the training patterns are linearly separable. @ 2003 Elsevier Ltd. All rights reserved. Neural networks have been widely used for solving supervised classification problems. In this paper, we consider the simplest feedforword neural network-the perceptron made up of m input neurons and one output neuron. The objective of training the neural networks is, for a given activation function g(x) : R1 + R1, to determine a weight vector W E R " , such that the training patterns {@}jJ=r are correctly classified according to the output C = g(W. f$) (cf. (2)). Some algorithms training the discrete perceptron where g(x) = sgn(x), such as the perceptron rule [l] and the delta rule (or Widrow-Hoff rule [2]) based on the LMS (least mean square) algorithm, have proved convergent for linearly separable training patterns. We are concerned in this paper with the continuous perceptron where g(x) is a sigmoidal function (a continuous function approximating the sign function sgn(x)). In th is case, the online gradient methods are often used for the network training, of which the convergence is our goal in this paper. We expect our analysis here can help to build up similar theories for.more important BP neural networks with hidden layers. In this respect, Gori and Maggini [3] prove a convergence result for BP neural networks with linearly separable patterns, under the assumption that the weight vectors keep bounded in the training process. We do not need this restriction in our case. To train the feedforward neural network (the perceptron), we are supplied with a set of training pattern pairs ([j, Oj}& c R " x {fl}, w h ere the ideal output Oj is " 1 " for a class, and "-1 "

منابع مشابه

Classification of linearly nonseparable patterns by linear threshold elements

Learning and convergence properties of linear threshold elements or perceptrons are well understood for the case where the input vectors (or the training sets) to the perceptron are linearly separable. Little is known, however, about the behavior of the perceptron learning algorithm when the training sets are linearly nonseparable. We present the first known results on the structure of linearly...

متن کامل

PhysRevE Convergence of stochastic learning in perceptrons with binary synapses

The efficacy of a biological synapse is naturally bounded, and at some resolution, latest at the level of single vesicles, it is discrete. The finite number of synaptic states dramatically reduce the storage capacity of a network when online learning is considered (i.e. the synapses are immediately modified by each pattern): the trace of old memories decays exponentially with the number of new ...

متن کامل

Internal representations of multi-layered perceptrons

Feedforward neural networks make incomprehensible decisions resulting from mappings learned from training examples defined in high dimensional feature spaces. What kind of internal representations are developed by multi-layered perceptrons (MLPs) and how do they change during training? Scatterograms of the training data transformed to the hidden and the output space reveal the dynamics of learn...

متن کامل

Convergence of stochastic learning in perceptrons with binary synapses.

The efficacy of a biological synapse is naturally bounded, and at some resolution, and is discrete at the latest level of single vesicles. The finite number of synaptic states dramatically reduce the storage capacity of a network when online learning is considered (i.e., the synapses are immediately modified by each pattern): the trace of old memories decays exponentially with the number of new...

متن کامل

On the Problem of Local Minima in Backpropagation

Supervised Learning in Multi-Layered Neural Networks (MLNs) has been recently proposed through the well-known Backpropagation algorithm. This is a gradient method which can get stuck in local minima, as simple examples can show. In this paper, some conditions on the network architecture and the learning environment are proposed which ensure the convergence of the Backpropagation algorithm. It i...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

متن کامل
عنوان ژورنال:
  • Appl. Math. Lett.

دوره 16  شماره 

صفحات  -

تاریخ انتشار 2003